Attention Mechanism-Based Light-Field View Synthesis

نویسندگان

چکیده

The angular information of light lost in conventional images but preserved and stored light-fields plays an instrumental role many applications such as depth estimation, 3D reconstruction post-capture refocusing. However, the limited resolution light-field due to consumer hardware limitations is a major drawback its widespread adoption. In this article, we present novel deep learning-based view synthesis method from sparse set input views. Our proposed method, end-to-end trainable network, utilizes convolutional block attention modules enhance built-in image-based rendering. module consists two sequentially applied channel spatial dimensions focus network on critical features. architecture combination three sub-networks, one for stereo feature extraction, another disparity last attention-based refinement. We schemes refinement that perform equally well differ number parameters by 1.2% execution time 44%. Quantitative qualitative results four challenging real-world datasets show superiority method. shows considerable PSNR gains, around 1 dB compared state art 0.5 over our previous LFVS-AM. addition, ablation studies demonstrate effectiveness each Finally, parallax edge precision-recall curve better preserves details.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Optimized Watermarking for Light Field Rendering based Free- View TV

In Free-View Television the viewers select freely the viewing position and angle of the transmitted multiview video. It is apparent that copyright and copy protection problems exist, since a video of this arbitrarily selected view can be recorded and then misused. In this context, the watermark in FTV should not only be resistant to common video processing and multi-view video processing operat...

متن کامل

Dynamic Field of View in a Tomographic Light Field Display

We also demonstrate the effectiveness of SART in generating dynamic real-time scenes (20 fps on an Nvidia Pascal GPU). Unlike traditional 2D displays, light field displays provide binocular depth cues such as high-resolution motion parallax, occlusion, translucency, and specularity. In particular attenuation-based light field displays provide an inexpensive means by which to produce these unenc...

متن کامل

Image-based View Synthesis

We present a new method for rendering novel images of exible 3D objects from a small number of example images in correspondence. The strength of the method is the ability to synthesize images whose viewing position is signi cantly far away from the viewing cone of the example images (\view extrapolation"), yet without ever modeling the 3D structure of the scene. The method relies on synthesizin...

متن کامل

Image Based View Synthesis

This dissertation deals with the image-based approach to synthesize a virtual scene using sparse images or a video sequence without the use of 3D models. In our scenario, a real dynamic or static scene is captured by a set of un-calibrated images from different viewpoints. After automatically recovering the geometric transformations between these images, a series of photo-realistic virtual view...

متن کامل

Reconstructing Dense Light Field from Multi-Focus Images Array for Virtual View Synthesis

This paper presents a novel method for synthesizing a novel view from two sets of differently focused images taken by an aperture camera array for a scene of two approximately constant depths. The proposed method consists of two steps. The first step is a view interpolation to reconstruct an all in-focus dense light field of the scene. The second step is to synthesize a novel view by light fiel...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Access

سال: 2022

ISSN: ['2169-3536']

DOI: https://doi.org/10.1109/access.2022.3142949